Parallelized hybrid optimization methods for nonsmooth problems using NOMAD and linesearch
نویسندگان
چکیده
Two parallelized hybrid methods are presented for single-function optimization problems with side constraints. The optimization problems are difficult not only due to possible existence of local minima and nonsmoothness of functions, but also due to the fact that objective function and constraint values for a solution vector can only be obtained by querying a black box whose execution requires considerable computational effort. Examples are optimization problems in Engineering where objective function and constraint values are computed via complex simulation programs, and where local minima exist and smoothness of functions is not assured. The hybrid methods consist of the well-known method NOMAD and two new methods called DENCON and DENPAR that are based on the linesearch scheme CS-DFN. The hybrid methods compute for each query a set of solution vectors that are evaluated in parallel. The hybrid methods have been tested on a set of difficult optimization problems produced by a certain seeding scheme for multiobjective optimization. We compare computational results with solution by NOMAD, DENCON, and DENPAR as stand-alone methods. It turns out that among the stand-alone methods, NOMAD is significantly better than DENCON and DENPAR. However, the hybrid methods are definitely better than NOMAD.
منابع مشابه
A Linesearch-Based Derivative-Free Approach for Nonsmooth Constrained Optimization
In this paper, we propose new linesearch-based methods for nonsmooth constrained optimization problems when first-order information on the problem functions is not available. In the first part, we describe a general framework for bound-constrained problems and analyze its convergence towards stationary points, using the Clarke-Jahn directional derivative. In the second part, we consider inequal...
متن کاملAn efficient one-layer recurrent neural network for solving a class of nonsmooth optimization problems
Constrained optimization problems have a wide range of applications in science, economics, and engineering. In this paper, a neural network model is proposed to solve a class of nonsmooth constrained optimization problems with a nonsmooth convex objective function subject to nonlinear inequality and affine equality constraints. It is a one-layer non-penalty recurrent neural network based on the...
متن کاملA Globally Convergent LP-Newton Method
We develop a globally convergent algorithm based on the LP-Newton method, which has been recently proposed for solving constrained equations, possibly nonsmooth and possibly with nonisolated solutions. The new algorithm makes use of linesearch for the natural merit function and preserves the strong local convergence properties of the original LP-Newton scheme. We also present computational expe...
متن کاملOptimality conditions for Pareto efficiency and proper ideal point in set-valued nonsmooth vector optimization using contingent cone
In this paper, we first present a new important property for Bouligand tangent cone (contingent cone) of a star-shaped set. We then establish optimality conditions for Pareto minima and proper ideal efficiencies in nonsmooth vector optimization problems by means of Bouligand tangent cone of image set, where the objective is generalized cone convex set-valued map, in general real normed spaces.
متن کاملNonsmooth Differential-algebraic Equations
A nonsmooth modeling paradigm for dynamic simulation and optimization of process operations is advocated. Nonsmooth differential-algebraic equations (DAEs) naturally model a wide range of physical systems encountered in chemical engineering conventionally viewed as exhibiting hybrid continuous/discrete behavior. Due to recent advancements in nonsmooth analysis, nonsmooth DAEs now have a suitabl...
متن کامل